EvoClass
AI012
Deep Dive into Large Language Models
Mainstream LLM Case Studies and Deployment Strategies
Learning Objectives
- Analyze the structural differences between Encoder-only (BERT), Decoder-only (GPT), and Encoder-Decoder (T5) architectures.
- Explain the three-stage training process: Pre-training (Base model), Instruction Tuning (SFT), and Alignment (RLHF/PPO).
- Compare the performance, scaling laws, and architectural innovations of mainstream LLMs including GPT, Llama, Qwen, and DeepSeek.